Lecture 17 | Sequence To Sequence: Attention Models